multitask - определение. Что такое multitask
Diclib.com
Словарь ChatGPT
Введите слово или словосочетание на любом языке 👆
Язык:

Перевод и анализ слов искусственным интеллектом ChatGPT

На этой странице Вы можете получить подробный анализ слова или словосочетания, произведенный с помощью лучшей на сегодняшний день технологии искусственного интеллекта:

  • как употребляется слово
  • частота употребления
  • используется оно чаще в устной или письменной речи
  • варианты перевода слова
  • примеры употребления (несколько фраз с переводом)
  • этимология

Что (кто) такое multitask - определение

FORM OF MACHINE LEARNING WHERE A MODEL LEARNS MULTIPLE TASKS
Applications of multi-task learning; Multitask learning
Найдено результатов: 9
Multi-task learning         
Multi-task learning (MTL) is a subfield of machine learning in which multiple learning tasks are solved at the same time, while exploiting commonalities and differences across tasks. This can result in improved learning efficiency and prediction accuracy for the task-specific models, when compared to training the models separately.
multitasking         
<computer, parallel> (Or "multi-tasking", "multiprogramming", "concurrent processing", "concurrency", "process scheduling") A technique used in an operating system for sharing a single processor between several independent jobs. The first multitasking operating systems were designed in the early 1960s. Under "cooperative multitasking" the running task decides when to give up the CPU and under "pre-emptive multitasking" (probably more common) a system process called the "scheduler" suspends the currently running task after it has run for a fixed period known as a "time-slice". In both cases the scheduler is responsible for selecting the next task to run and (re)starting it. The running task may relinquish control voluntarily even in a pre-emptive system if it is waiting for some external event. In either system a task may be suspended prematurely if a hardware interrupt occurs, especially if a higher priority task was waiting for this event and has therefore become runnable. The scheduling algorithm used by the scheduler determines which task will run next. Some common examples are round-robin scheduling, priority scheduling, {shortest job first} and guaranteed scheduling. Multitasking introduces overheads because the processor spends some time in choosing the next job to run and in saving and restoring tasks' state, but it reduces the worst-case time from job submission to completion compared with a simple batch system where each job must finish before the next one starts. Multitasking also means that while one task is waiting for some external event, the CPU to do useful work on other tasks. A multitasking operating system should provide some degree of protection of one task from another to prevent tasks from interacting in unexpected ways such as accidentally modifying the contents of each other's memory areas. The jobs in a multitasking system may belong to one or many users. This is distinct from parallel processing where one user runs several tasks on several processors. Time-sharing is almost synonymous but implies that there is more than one user. Multithreading is a kind of multitasking with low overheads and no protection of tasks from each other, all threads share the same memory. (1998-04-24)
multitasking         
¦ noun Computing the execution of more than one program or task simultaneously by sharing the resources of the computer processor between them.
Derivatives
multitask verb
Media multitasking         
USING TV, THE WEB, RADIO, TELEPHONE, PRINT, OR ANY OTHER MEDIA IN CONJUNCTION WITH ANOTHER
Media Multitasking
Media multitasking is the concurrent use of multiple digital media streams. Media multitasking has been associated with depressive symptoms and social anxiety by a single study involving 318 participants.
cooperative multitasking         
<parallel, operating system> A form of multitasking where it is the responsibility of the currently running task to give up the processor to allow other tasks to run. This contrasts with pre-emptive multitasking where the task scheduler periodically suspends the running task and restarts another. Cooperative multitasking requires the programmer to place calls at suitable points in his code to allow his task to be descheduled which is not always easy if there is no obvious top-level main loop or some routines run for a long time. If a task does not allow itself to be descheduled all other tasks on the system will appear to "freeze" and will not respond to user action. The advantage of cooperative multitasking is that the programmer knows where the program will be descheduled and can make sure that this will not cause unwanted interaction with other processes. Under pre-emptive multitasking, the scheduler must ensure that sufficient state for each process is saved and restored that they will not interfere. Thus cooperative multitasking can have lower overheads than pre-emptive multitasking because of the greater control it offers over when a task may be descheduled. Cooperative multitasking is used in RISC OS, {Microsoft Windows} and Macintosh System 7. (1995-03-20)
Human multitasking         
ABILITY TO PERFORM MORE THAN ONE ACTIVITY AT THE SAME TIME
Human Multitasking
Human multitasking is the concept that one can split their attention on more than one task or activity at the same time, such as speaking on the phone while driving a car. Multitasking can result in time wasted due to human context switching and becoming prone to errors due to insufficient attention.
Cooperative multitasking         
Cooperative multitasking, also known as non-preemptive multitasking, is a style of computer multitasking in which the operating system never initiates a context switch from a running process to another process. Instead, in order to run multiple applications concurrently, processes voluntarily yield control periodically or when idle or logically blocked.
Multitask optimization         
Multi-task optimization is a paradigm in the optimization literature that focuses on solving multiple self-contained tasks simultaneously.Gupta, A.
pre-emptive multitasking         
ACT OF TEMPORARILY INTERRUPTING A TASK BEING CARRIED OUT BY A COMPUTER SYSTEM, WITHOUT REQUIRING ITS COOPERATION, AND WITH THE INTENTION OF RESUMING THE TASK AT A LATER TIME
Pre-emptive multitasking; Preemptive multitasking; Pre-emptive multi-tasking; Preemptive multithreading; Time slice; Pre-emptive multithreading; Pre-emptive multitasking operating system; Preemptive scheduler; Pre-emptive scheduler; Preemptive scheduling
<operating system, parallel> A type of multitasking where the scheduler can interrupt and suspend ("swap out") the currently running task in order to start or continue running ("swap in") another task. The tasks under pre-emptive multitasking can be written as though they were the only task and the scheduler decides when to swap them. The scheduler must ensure that when swapping tasks, sufficient state is saved and restored that tasks do not interfere. The length of time for which a process runs is known as its "time slice" and may depend on the task's priority or its use of resources such as memory and I/O. OS/2, Unix and the Amiga use pre-emptive multitasking. This contrasts with cooperative multitasking where each task must include calls to allow it to be descheduled periodically. (1995-03-20)

Википедия

Multi-task learning

Multi-task learning (MTL) is a subfield of machine learning in which multiple learning tasks are solved at the same time, while exploiting commonalities and differences across tasks. This can result in improved learning efficiency and prediction accuracy for the task-specific models, when compared to training the models separately. Early versions of MTL were called "hints".

In a widely cited 1997 paper, Rich Caruana gave the following characterization:

Multitask Learning is an approach to inductive transfer that improves generalization by using the domain information contained in the training signals of related tasks as an inductive bias. It does this by learning tasks in parallel while using a shared representation; what is learned for each task can help other tasks be learned better.

In the classification context, MTL aims to improve the performance of multiple classification tasks by learning them jointly. One example is a spam-filter, which can be treated as distinct but related classification tasks across different users. To make this more concrete, consider that different people have different distributions of features which distinguish spam emails from legitimate ones, for example an English speaker may find that all emails in Russian are spam, not so for Russian speakers. Yet there is a definite commonality in this classification task across users, for example one common feature might be text related to money transfer. Solving each user's spam classification problem jointly via MTL can let the solutions inform each other and improve performance. Further examples of settings for MTL include multiclass classification and multi-label classification.

Multi-task learning works because regularization induced by requiring an algorithm to perform well on a related task can be superior to regularization that prevents overfitting by penalizing all complexity uniformly. One situation where MTL may be particularly helpful is if the tasks share significant commonalities and are generally slightly under sampled. However, as discussed below, MTL has also been shown to be beneficial for learning unrelated tasks.